首页> 外文OA文献 >Asynchronous Stochastic Approximation Based Learning Algorithms for As-You-Go Deployment of Wireless Relay Networks along a Line
【2h】

Asynchronous Stochastic Approximation Based Learning Algorithms for As-You-Go Deployment of Wireless Relay Networks along a Line

机译:基于异步随机逼近的maTLaB学习算法   随时随地部署无线中继网络

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

We are motivated by the need for impromptu (or as-you-go) deployment ofmultihop wireless networks, by human agents or robots; the agent moves along aline, makes wireless link quality measurements at regular intervals, and makeson-line placement decisions using these measurements. As a first step, we haveformulated such deployment along a line as a sequential decision problem. Inour earlier work, we proposed two possible deployment approaches: (i) the pureas-you-go approach where the deployment agent can only move forward, and (ii)the explore-forward approach where the deployment agent explores a fewsuccessive steps and then selects the best relay placement location. The latterwas shown to provide better performance but at the expense of more measurementsand deployment time, which makes explore-forward impractical for quickdeployment by an energy constrained agent such as a UAV. Further, thedeployment algorithm should not require prior knowledge of the parameters ofthe wireless propagation model. In [1] we, therefore, developed learningalgorithms for the explore-forward approach. The current paper provides deploy-and-learn algorithms for the pure as-you-goapproach. We formulate the sequential relay deployment problem as an averagecost Markov decision process (MDP), which trades off among power consumption,link outage probabilities, and the number of deployed relay nodes. First weshow structural results for the optimal policy. Next, by exploiting the specialstructure of the optimality equation and by using the theory of asynchronousstochastic approximation, we develop two learning algorithms thatasymptotically converge to the set of optimal policies as deploymentprogresses. Numerical results show reasonably fast speed of convergence, andhence the model-free algorithms can be useful for practical, fast deployment ofemergency wireless networks.
机译:我们受到人工代理或机器人对即兴部署多跳无线网络的需求的激励。代理沿着一条线移动,定期进行无线链路质量测量,并使用这些测量结果做出子线布置决策。第一步,我们将这种部署作为顺序决策问题沿线进行了规划。在我们的早期工作中,我们提出了两种可能的部署方法:(i)部署代理只能向前移动的“即取即用”方法,以及(ii)部署代理先探索一些成功步骤然后选择的“探索-前进”方法。最佳的继电器放置位置。后者被证明可以提供更好的性能,但以更多的测量和部署时间为代价,这使得通过能量受限的代理(例如UAV)进行快速部署变得不切实际。此外,部署算法应该不需要无线传播模型的参数的先验知识。因此,在[1]中,我们开发了探索算法的学习算法。当前的文章提供了纯学习方法的部署和学习算法。我们将顺序中继部署问题公式化为平均成本马尔可夫决策过程(MDP),该过程在功耗,链路中断概率和已部署中继节点的数量之间进行权衡。首先,我们显示最佳策略的结构结果。接下来,通过利用最优性方程的特殊结构,并使用异步随机逼近理论,我们开发了两种学习算法,它们随着部署进度渐近收敛于最优策略集。数值结果表明收敛速度相当快,因此,无模型算法对于应急无线网络的实用,快速部署很有用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号